专利摘要:
method and arrangement for video encoding methods and arrangements in video encoding and decoding entities. the methods and arrangements refer to splicing coding of reference information associated with encoded video. the method and arrangement in a decoding entity refer to obtaining (402) a unique syntax element associated with a coded block be, and identifying (404) a reference mode and one or more based reference pictures in the obtained syntax element. the method and arrangement further refer to decoding (406) the b-block based on the identified reference mode and one or more reference pictures, thus providing a decoded block, b, of pixels.
公开号:BR112012015001B1
申请号:R112012015001-4
申请日:2010-12-17
公开日:2021-06-29
发明作者:Zhuangfei Wu;Kenneth Andersson;Clinton Priddle;Thomas Rusert;Rickard Sjoberg
申请人:Telefonaktiebolaget Lm Ericsson [Publ];
IPC主号:
专利说明:

Technical Field
[001] The invention generally relates to a method and an arrangement for encoding information related to video encoding. Fundamentals
[002] Video encoding can be performed in intra mode and/or inter mode. Intra mode exploits redundancies within a video frame, and inter mode exploits redundancies between video frames. In inter mode, pixel luminance/chroma predictions are obtained from already encoded/decoded images called reference images. Depending on the number of reference images used for prediction, inter mode is categorized into uni-prediction mode (or uni-direction mode), bi-prediction mode (B mode), and possibly tri-prediction mode, etc. , where, respectively, 1, 2 and 3 reference images are used. Within the documents, these different modes, ie, uni-forecast, bi-forecast, etc., will be referred to as “reference modes”.
[003] Advanced Video Coding (AVC), which is also known as H.264 and MPEG-4 Part 10, is the state of the art standard for 2D video coding from ITU-T (International Telecommunication Union - Sector Telecommunication Standardization) and MPEG (Moving Image Specialist Group). The AVC codec is a hybrid codec, which takes advantage of eliminating redundancy between frames and within a frame.
[004] In AVC, indicators of the relevant reference images are placed in order in two reference lists. Indicators are denoted benchmarks, and are numbered from 0 to N, for example (0.1,...,N). The first list, List 0 (L0), mainly manages past reference images, that is, reference images prior to an image current in time, and the second list, List 1 (L1), typically manages future reference images , that is, reference images subsequent to an image current in time. For low delay video encoding, L1 can also manage past reference pictures. Each list can contain indexes of up to 15 reference images, that is, N=14.
[005] Also, in AVC, an indicator, or reference mode index, specifying the selection of one of the reference image lists (for example, for uni-forecast), or both reference image lists (by example, for bi-prediction), is encoded along with the partition structure in Block Macro (MB) mode/sub-MD mode, while the indicators, or reference image indices, specifying the reference images selected in the respective lists are encoded as separate syntax elements. “Partition structure” refers to partitions, such as, for example, 16x16, 16x8 or 8x16, of a 16x16 MB. A partition, for example 16x16, is typically associated with a motion vector (MV) and a reference index when uni-forecast is used, and with two MVs and two reference indices when bi-forecast is used. A MV has a horizontal component MVx and a vertical component MVy which describes how pixels of the current partition are produced from the corresponding reference image, such as Ipred(x,y)=Iref(x-MVx,y-MVy) .
[006] The number of reference images associated with an image or partition depends on the reference mode associated with the same partition, ie, whether it is uni-predict or bi-predict, etc. When decoding reference information in a decoder, both the reference mode indices and one or more reference picture indices associated with a picture or partition must be correctly decoded in order for the decoder to be able to decode the image or partition correctly. Incorrect decoding of any of the reference mode indices and one or more reference image indices can result in misinterpretations of the reference information.
[007] Current reference information encoding methods, such as the AVC method described above, require a relatively large number of bits in order to transmit the reference information associated with each block. This is identified as inefficient in terms of coding efficiency. summary
[008] It would be desirable to allow improved coding efficiency for reference information, i.e., information identifying one or more reference pictures used for prediction of a current frame. It is an object of the invention to enable improved coding efficiency for reference information. Furthermore, it is an object of the invention to provide a method and an arrangement to enable an improved encoding of reference information. These objectives can be met by a method and an arrangement in accordance with the appended independent claims. Optional modalities are defined by the dependent claims. The prediction, encoding and decoding described below can be performed within the same entity or node, or on different entities or nodes.
[009] According to a first aspect, a method is provided in a video decoding entity. The method comprises obtaining a single syntax element associated with a Be encoded block. The method further comprises identifying a reference mode and one or more reference pictures based on the obtained syntax element, and decoding the Be block, based on the identified reference mode and one or more reference pictures, thus providing a block decoded, B, of pixels.
[0010] According to a second aspect, an array is provided in a video decoding entity. The arrangement comprises a functional unit, which is adapted to obtain a single syntax element associated with a coded block Be. The arrangement further comprises the functional unit, which is adapted to identify a reference mode and one or more reference pictures based on the obtained syntax element. The arrangement further comprises a functional unit, which is adapted to decode block B on the basis of the identified reference mode and one or more reference pictures, thus providing a decoded block B of pixels.
[0011] According to a third aspect, a method is provided in a video encoding entity. The method comprises encoding a B-block of pixels using a reference mode and one or more reference pictures, thereby providing a B-coded block. The method further comprises deriving a single syntax element, identifying the reference mode and one or more reference pictures(s) and providing the single syntax element to a Be block decoder.
[0012] According to a fourth aspect, an array is provided in a video encoding entity. The arrangement comprises a functional unit, which is adapted to encode a B-block of pixels using a reference mode and one or more reference pictures, thus providing a B-coded block. The arrangement further comprises a functional unit, which is adapted to derive a single syntax element identifying the mode of reference and one or more reference pictures. The arrangement further comprises a functional unit, which is adapted to provide the single syntax element to a Be block decoder.
[0013] The above methods and provisions can be used to improve coding efficiency. Coding efficiency can be improved, due to the fact that the use of fewer bits in order to identify one or more reference pictures used for prediction of a current frame is enabled. The above methods and arrangements can further allow for improved error resilience performance. Furthermore, by grouping reference information to form a single syntax element, easy manipulation of index reference numbers becomes possible. Furthermore, the use of a syntax element as described above will allow the use of which some combinations of reference index and reference mode are more likely than others, allowing efficient encoding of these combinations. For example, smaller codewords could be assigned to more likely combinations of reference index and reference mode.
[0014] The above methods and provisions can be implemented in different modalities. In some embodiments, the identification of a reference mode and one or more reference images is based on a predefined mapping between the syntax element and the reference mode and one or more specific reference images. In some embodiments, the single syntax element represents an entry in a predefined first reference list, which may comprise one or more entries. An input may identify a plurality of reference images or a single reference image, and may also further identify a reference mode.
[0015] The single syntax element may further represent a reference mode and an entry in a second predefined reference list, which may comprise one or more entries identifying a single reference image, respectively.
[0016] Entries in lists can be identified by list indices. In addition, the number of bits representing the syntax element obtained can be related to the probability of the syntax element-specific values.
[0017] In some embodiments, prediction of reference information can be performed for Be (or B when in entity encoding), based on the unique syntax elements associated with neighboring blocks of Be (or B). Furthermore, in some embodiments, subregions of a block associated with multi-forecast can be identified, so that the subregions of the respective corresponding regions of the multi-forecast reference blocks have a relatively low correlation between them, and then a alternative forecasting, rather than multi-forecasting, can be used for the identified subregions.
[0018] The above modalities have mainly been described in terms of a method. However, the above description is also intended to encompass modalities of the arrangements, adapted to enable the performance of the characteristics described above. The different features of the above exemplary modalities can be combined in different ways according to need, requirement or preference.
[0019] According to yet another aspect, a computer program is provided, which comprises computer-readable code means, which when executed in one or more processing units, cause any of the provisions described above to perform the corresponding procedure of according to one of the methods described above.
According to yet another aspect, a computer program product is provided which comprises the above computer program. Brief Description of Drawings
[0021] The invention will now be described in greater detail by means of exemplary embodiments and with reference to the accompanying drawings, in which: Figure 1 is a schematic view illustrating a conversion from a representation of reference information according to the technique above for a representation of the reference information according to an exemplary embodiment.
[0022] Figure 2 is a table showing differences between an AVC reference index representation and a reference index representation according to an exemplary embodiment.
[0023] Figure 3 is a schematic view illustrating the designation of the indicator reference information in accordance with an exemplary embodiment.
[0024] Figure 4 is a flowchart illustrating a procedure for decoding the set of encoded information related to a reference mode and one or more reference images in a video decoding entity, according to an exemplary embodiment.
[0025] Figure 5 is a block diagram illustrating an adapted arrangement for decoding the set of encoded information related to a reference mode and one or more reference pictures in a video decoding entity, according to an exemplary embodiment.
[0026] Figure 6 is a flowchart illustrating a procedure for joining coding information related to a reference mode and one or more reference pictures into a video coding entity, according to an exemplary embodiment.
[0027] Figure 7 is a block diagram illustrating an arrangement adapted for coding joining information related to a reference mode and one or more reference pictures in a video coding entity, according to an exemplary embodiment.
[0028] Figure 8 is a schematic view illustrating an arrangement at a video encoding/decoding input, according to an exemplary embodiment.
[0029] Figures 9 and 10 are schematic views illustrating the determination of the frequency of occurrence of different combinations of a reference mode and one or more reference images associated with neighboring blocks of a current block, according to exemplary embodiments.
[0030] Figure 11 is a schematic view illustrating designation of indicators (code words) for different index symbols, according to the prior art.
[0031] Figure 12 is a schematic view illustrating designation of indicators (code words), according to an exemplary embodiment.
[0032] Figure 13 is a schematic view illustrating partition based on implicit information, according to an exemplary embodiment. Detailed Description
[0033] Briefly described, a new procedure for representing and transmitting reference information, i.e. reference mode(s) and reference pictures are provided for the interprediction of encoding and decoding. The procedure may be referred to as Benchmark Signaling, or Benchmark Information Indicator Signaling (RIS).
[0034] Within this document, the term "neighbor blocks of block X" is used as a reference to blocks that are neighbors to block X, that is, located neighbors to or in the vicinity of block X. Also, within this document, the term “block” is used to refer to a unit of pixels. The term “reference image” or “reference block” is used to refer to a previously encoded/decoded image, a block, a region or an area of an image, this image, block, region, etc., used as a reference for the forecast.
[0035] When using RIS, instead of, for example, encoding a reference mode indicator in close association with a partition structure indicator, and encoding reference image indicators separately, for example, as in AVC, the indicator of the reference mode and the reference picture pointers associated with a coded block are "packed together in one place", i.e. they are coded together. Coding the union of reference mode indicators and reference images, i.e. reference information, results in that a single syntax element, or indicator, represents all information in the necessary reference modes and reference images in order to decode the encoded block in a satisfactory manner. That is, since this unique syntax element is given for an encoded block, a decoder must be able to identify the reference pictures required for decoding the block. The “syntax element” may also be denoted, for example, “syntax unit”, “joint indication unit” or “joint identification unit”.
[0036] One way to describe RIS is to describe a "conversion" or mapping from a traditional representation, such as, for example, the AVC representation of the reference information using two separate lists, to an illustrative representation of the reference information of according to RIS. Such a conversion to RIS representation could basically be done in three steps, as illustrated in Figure 1.
[0037] The first step 102 could be to form a single reference index list from multiple index lists. For example, instead of managing two reference index lists as in AVC, all reference image indexes can be sorted in a certain order into a single joint list, as an alternative or a complement to two AVC lists. This is illustrated in Figure 1, where the L0 and L1 picture reference index lists are merged, or multiplexed, into a new LRIS list, in an interleaved manner. Furthermore, in a second step, 104, the index numbers can be reassigned accordingly, to follow a consecutive order, ie, 0-5 in the new LRIS list.
[0038] The index numbers, or entries, in the LRIS list after step 104 represent information regarding both a reference mode (uni-prediction backwards or forwards) and a reference image. An index to an entry in the LRIS can be denoted, for example, a “RIS index” or an “index parameter”. The RIS index numbers 0-5 in LRIS, after step 104 in this example, represent uni-prediction from four past images (originally at L0 = (0,1,2,3)), and two future images (originally in L1=(0.1)).
[0039] In addition, one or more list entries representing bi-forecast can be added to LRIS, for example, by inserting or appending. Thus, RIS indices indicative of inputs representing bi-prediction do not point to a single reference image, but rather two reference images. Thus, a RIS index can identify a combination of a reference mode and one or more reference images.
[0040] Consequently, in a final step 106, inputs related to bi-prediction mode, where two reference images are used for prediction, can be consecutively appended to LRIS, and be indicated or represented by RIS indices. For example, the entry with RIS index number 7 can be configured to signal or suggest that the current image is using image number 0 and image number 1 as bi-prediction references. Thus, this information is inherent in RIS index 7. Index number 8 can similarly be configured to suggest that the current image is using image number 0 and image number 2 as bi-prediction references. Analogously, the LRIS list can be further extended with entries representing tri-prediction, identifying three reference images, and so on.
[0041] Alternatively, steps 104 and 106 can be performed in reverse order such that entries related to the bi-prediction mode are added first, eg inserted or appended, and then the index numbers are reassigned accordingly. As previously described, entries related to bi-forecast mode could also be inserted, for example, between entries related to uni-forecast, which would require that index number reassignment be performed after insertion, as a complement or alternative to step 104. In this example, the mapping is represented by a single reference list, of which the indices of the different inputs represent a reference mode and one or more reference images. It should be noted that this is just an optional example, and that the mapping may involve several steps, and that no express list or record of the exemplified type is a requirement for performing the mapping.
[0042] An example of the difference between an AVC reference index representation and a RIS index representation, according to an example modality, is shown in a table in Figure 2. In this example, it is assumed that there are four reference images available for encoding a current picture, of which two reference pictures are past reference pictures and two are future reference pictures. In this exemplary RIS representation, indices 0, 1, 3, and 4 are set to indicate uni-prediction from a respective one of the four reference images. Indices 2 and 5 are defined to indicate bi-prediction from a respective pair of four reference images. It should be noted that a reference index AVC signaling would also comprise information related to partitions, as this information is encoded together with the reference index mode, such as, for example, “INTER_16x16_L0”. This is not, however, shown in figure 2.
[0043] In the example shown in the table in Figure 2, some of the RIS indices indicating or representing bi-forecast are placed immediately after the "closest" uni-forecast RIS indices, that is, interspersed with the indices representing uni-forecast. This representation of the RIS index is further illustrated in Figure 3, which shows a so-called Hierarchical Image Group 7B (BGOP). In the figure, the so-called “current frame”, that is, the frame to be encoded, is frame 3 in GOP 7B. The RIS indices shown in Figure 3 correspond to the RIS indices 0-7 in the table in Figure 2. An alternative RIS representation could be to let the RIS indices 0-3 indicate uni-forecast, and the following RIS indices indicate bi-forecast, as in example illustrated in figure 1.
[0044] The ways to define the meaning of a RIS index, or RIS parameter, are not limited by the example given in this document. For example, a mathematical formula could be defined to interpret the meaning of the RIS index, for example, a function with 2 variables f (RIS_index, current_frame_num) that returns the identification of 2 reference image indices for a bi-forecast RIS index and identifies a reference image index to a one-way RIS index, and so on. In one example, chain_frame_num corresponds to the frame number within a 7B-picture BGOP, where 0 is the first frame in display order and 8 is the last frame in the BGOP. In another example, RIS index is always designated using the formula:


[0045] Where refidx0 and refidx1 are the indices in the reference list L0 and L1 respectively. L0_len and L1_len are the length of list L0 and L1 respectively.
[0046] Alternatively, a table can be used to match the RIS index with two corresponding unidirectional indexes in the case of bi-forecasting and a unidirectional index in the case of a single prediction. Which method to select depends, for example, on hardware/software restrictions.
[0047] However, regardless of which method is used for deriving a syntax element, the method must be known by both the encoder and the decoder, such that the encoder is enabled to derive and provide a correct syntax element, and the decoder is enabled to interpret the syntax element correctly and thus identify the reference information necessary to decode the coded block or frame in question.
[0048] The RIS index can apply to different levels of video encoding, for example, frame level, higher MB level, MB level or sub MB level. Example procedure, figure 4, decoding
[0049] An embodiment of the decoding part of the reference information transmission procedure will now be described with reference to Figure 4. The procedure could be performed on a video decoding entity, which could be a video decoder or an entity comprising still functional units plus a video decoder. Initially, a single syntax element associated with a Be encoded block is obtained in an action 402. The single syntax element can be a unit, eg a symbol, in "the bit stream", ie the encoded representation of, for example, a video sequence, or being a unit, which is decoded from the bitstream. The syntax element is one or more bits representing a number that corresponds to reference information, such as, for example, a RIS index. Typically, fewer bits are used to represent RIS indices that are relatively common compared to the number of bits used to represent RIS indices that are less common. The syntax element is decoded from the bit stream to obtain the number, eg RIS index, that it represents. Decoding can be done according to VLC (Variable Length Coding) or arithmetic coding such as CABAC (Context Adapted Binary Arithmetic Coding), for example.
[0050] Then, in a 404 action, a reference mode and one or more reference images to be used when decoding the Be block are identified based on the obtained syntax element. The identified reference mode and one or more reference pictures correspond to the reference mode and pictures used when encoding the block in an encoder. Identification may involve, for example, demapping, deciphering or "decoding" the syntax element using a mapping table, a reference list or other predefined information or function, by the use of which a reference mode and one or more images of reference can be identified, given an element of syntax. Furthermore, when having identified the reference mode and one or more required reference pictures, the coded block Be, which is assumed to be obtained using conventional methods, is decoded in an action 406.
[0051] The single syntax element may be an indicator or index, for example, denoted RIS index, of an entry in a reference list, which reference list may comprise a plurality of entries, each entry representing or identifying one or more reference modes and one or more reference images. Alternatively, the syntax element is a code word corresponding to an entry in a lookup table. The lookup table may link the code word, for example, to a reference mode and one or more entries in one or more reference list, such as, for example, L0 and L1 in AVC. The reference mode can define which single reference list or multiple reference lists are to be used in block decoding. Layout example, figure 5, decoding
[0052] Below, an example of the arrangement 500, adapted to enable the performance of the above-described decoding procedure, will be described with reference to Figure 5. The arrangement is illustrated as being located in a video decoding entity, 501, which it could be a video decoder or an entity further comprising a functional unit in addition to a video decoder, such as, for example, a computer, a mobile terminal or a dedicated video device. Arrangement 500 is further illustrated for communicating with other entities through a communication unit 502, which can be considered to comprise conventional means for any type of wired or wireless communication. Encoded video to be decoded is assumed to be obtained from the communication unit 502 or a memory by an acquisition unit 504, and encoded blocks are assumed to be decoded in a decoding unit 508, where the functional unit 508 uses methods of conventional decoding.
[0053] The obtaining unit 504 is adapted to obtain a single syntax element associated with a Be coded block. Arrangement 500 further comprises an identification unit 506, which is adapted to identify a reference mode and one or more reference pictures to be used when decoding block Be, based on the obtained syntax element. As previously described, the arrangement 500 further comprises a decoding unit 508, which is adapted to decode block B based on the determined mode of reference and reference pictures, thus providing a decoded block B of pixels.
[0054] In this arrangement, the syntax element may be an indicator or index of an entry in a reference list, reference list may comprise a plurality of entries, each entry representing or identifying one or more modes of reference and one or more reference images. Alternatively, the arrangement can be adapted to another case when the syntax element is a codeword corresponding to an entry in a lookup table. The lookup table can link the code word, for example, to a reference mode and one or more entries in one or more reference lists, such as, for example, L0 and L1 in AVC.
[0055] The video decoding entity 501 may further comprise, for example, a display unit 510 adapted to display the decoded video. Example procedure, figure 6, coding
[0056] An embodiment of the encoding part of the reference information transmission procedure will now be described with reference to Figure 6. The procedure could be performed on a video encoding entity, which could be a video encoder, or a entity further comprising a functional unit in addition to a video encoder. Initially, a B-block of pixels is encoded in an action 602 using a reference mode and one or more reference images, thus providing a Be-encoded block.
[0057] Then, a single syntax element is derived, in an action 604, based on the reference mode and one or more reference images used for encoding, the syntax element thus identifying, directly or indirectly, the reference mode and the one or more reference images used for encoding block B. The syntax element could, for example, be derived by locating a list entry, corresponding to the reference mode and reference images used, in a predefined reference list , and then defining the index number of said entry to constitute the syntax element. Alternatively, a predefined mapping table or lookup table could provide a mapping between different combinations of reference modes and reference images and different syntax elements. The syntax element could also be an argument to a predefined function, a function that returns a reference mode indicator and one or more reference image indicators. Such an "argument" syntax element could be derived, for example, by means of a predefined "reverse function", taking a reference mode indicator and one or more reference image indicators as arguments and returning a single syntax element.
[0058] In addition, the derived syntax element is provided to a Be block decoder, in association with the Be block, in an action 606. Thus, the reference information, that is, the information in the reference mode and in the one or more reference pictures used when encoding the B block, also to be used when decoding the B encoded block, can be transmitted to a decoder in a compact and error resilient manner. The syntax element could, for example, be provided by being transmitted over a radio channel to an entity or node comprising a decoder. Furthermore, the syntax element could, for example, be stored in memory along with the associated encoded video and be accessed by a decoding entity at another point in time. Layout example, figure 7, coding
[0059] Below, an example arrangement 700, adapted to enable performance of the above-described procedure related to encoding, will be described with reference to Figure 7. The arrangement is illustrated as being located in a video encoding entity, 701, which could be a video encoder or an entity further comprising functional units in addition to a video encoder, such as, for example, a computer, a mobile terminal or a dedicated video device. Arrangement 700 can communicate with other entities through a communication unit (not shown), which can be considered to comprise conventional means for any type of wired or wireless communication. Decoded video to be encoded is assumed to be obtained, for example, from the communication unit or a memory.
[0060] Arrangement 700 comprises a coding unit 702, which is adapted to encode a block, B, of pixels using a reference mode and one or more reference pictures, thus providing a coded block Be. Arrangement 700 further comprises a derivation unit 704 which is adapted to derive a single syntax element which directly or indirectly identifies the mode of reference and the one or more reference pictures used when encoding block B. index element could be derived in different ways, as previously described, and could be, for example, an indicator, such as, for example, an index, or a codeword, etc.
[0061] Arrangement 700 further comprises a supply unit 706, which is adapted to supply the single syntax element to a block decoder Be, possibly via a communication unit. The single syntax element can be provided, for example, by transmission over a radio channel to an entity or node comprising a decoder. Layout example, figure 8
[0062] Figure 8 schematically shows an embodiment of an arrangement 800 in a video decoding entity, which may also be an alternative way of disclosing an embodiment of the arrangement for decoding in a video decoding entity illustrated in Figure 5. Comprised of arrangement 800 here is a processing unit 806, for example with a DSP (Digital Signal Processor). Processing unit 806 can be a single unit or a plurality of units to perform different actions of procedures described herein. Arrangement 800 may also comprise an input unit 802 for receiving signals from other entities, and an output unit 804 for providing signals to other entities. Input unit 802 and output unit 804 can be arranged as an integrated entity.
[0063] Furthermore, the arrangement 800 comprises at least one computer program product 808 in the form of a non-volatile memory, for example, an EEPROM (Electrically Erasable Programmable Read Only Memory), a flash memory and a hard disk. The computer program product 808 comprises a computer program 810, which comprises code means, which when executed in the processing unit 806 in the arrangement 800 causes the arrangement and/or the video decoding entity to perform the actions of the procedures. described above in conjunction with Figure 4.
[0064] The computer program 810 may be configured as a computer program code structured in computer program modules. Accordingly, in the described exemplary embodiments, code means in computer program 810 of array 800 comprises an acquisition module 810a for obtaining a unique syntax element associated with an encoded video unit/block, e.g., by decoding the from a bit stream originating from a data transmission entity or from a storage, for example, a memory. The computer program further comprises an identification module 810b for identifying a mode of reference and one or more reference pictures based on the obtained syntax element. The computer program 810 further comprises a decoding module 810c for decoding the coded block.
[0065] The 810a-c modules could essentially perform the actions of the flow illustrated in Figure 4, to mimic the arrangement in a video decoding entity illustrated in Figure 5. In other words, when the different modules 810a-c are executed in the processing unit 806, they correspond to units 502-506 of figure 5.
[0066] Similarly, a corresponding alternative to the arrangement illustrated in Figure 7 is possible.
[0067] Although the code means in the modality described above in conjunction with Figure 8 are implemented as computer program modules which, when executed in the processing unit, cause the arrangement and/or video processing/presentation entity perform the actions described above in conjunction with the figure mentioned above, at least one of the code means can in alternative embodiments be implemented at least in part as hardware circuits.
[0068] The processor can be a single CPU (Central Processing Unit), but it could also comprise two or more processing units. For example, the processor may include general purpose microprocessors; instruction set processors and/or related chip sets and/or special purpose microprocessors such as ASICs (Application Specific Integrated Circuits). The processor may also comprise board memory for caching purposes. The computer program product may be carried by a computer program product connected to the processor. The computer program product comprises a computer readable medium on which the computer program is stored. For example, the computer program product may be flash memory, a RAM (Random Access Memory), ROM (Read Only Memory) or an EEPROM, and the computer program modules described above could, in alternative embodiments, be distributed in different computer program products in the form of memories within the receiving unit.
[0069] Furthermore, it is realized that a problem with existing video coding technology is that no satisfactory reference mode/reference image prediction scheme is defined or applied. Below, such a prediction scheme will be described. It is noticed that in, for example, a cut/fade/flash scene, it is quite common for the same, for example, combination of reference mode and reference images to be used for the prediction of neighboring or adjacent MBs. Furthermore, it is clear that current coding methods do not take advantage of the correlation between reference modes/reference images used for neighboring MBs. In previous solutions, separate identification components from the MB reference information are independently encoded and transmitted to a video decoder.
[0070] An exemplified prediction scheme at MB level could be described as follows. The exemplified prediction scheme applies to both encoder and decoder, and could be applied to any block size.
[0071] In association with the encoding/decoding of a current MB, the encoder/decoder is arranged to analyze the reference indicators of the surrounding encoded MBs, also called “the context” of the MB. These surrounding blocks could also be denoted the “neighbor blocks” of the current block. The encoder/decoder counts the number of times each out of a set of candidate indicators or indices appears between neighboring blocks, and selects one, for example, according to a predefined scheme, with the highest count, as a prediction or I estimated. The selected benchmark must be related to the inter forecast. The selected reference indicator is defined by a prediction or estimate of which reference images (and reference mode) that may be suitable for use when encoding/decoding the current MB. The prediction is derived by analyzing the information related to the neighboring encoded/decoded blocks of the MB, rather than analyzing the current MB of the same. In an encoder, the MB can either be encoded or decoded during this selection of a prediction, as in this example, the prediction is not to be used for selection of reference images (and reference mode) for use when encoding the Current MB. In a decoder, the current MB is encoded during prediction.
[0072] An exemplary neighbor block reference indicator analysis and selection of an estimate are illustrated in figure 9. In the example illustrated in figure 9, four neighbor blocks of a current block are considered. However, the method is also applied to other sets or subsets of neighboring blocks considered. An example of a set or neighboring blocks could consist of, for example, a block to the left, a block to the left at the top, and a block above the current block. Other set examples could comprise just the block on the left and the block above. In figure 9, the neighboring blocks of the current block are associated with the respective reference flags, or indices, 1, 1, 2 and 0. Thus, the reference flag "1" has the highest count, that is, the highest frequency of occurrence, as it appears twice between neighboring blocks. Thus, the reference indicator "1" is selected to represent the prediction or estimate of the reference images (and mode) used, or to be used, when encoding a current block, or, when the prediction occurs in a decoder, predicting the reference (and mode) images to be used when decoding the current block.
[0073] When counting the number of occurrences of a certain reference indicator among the neighboring blocks, more than one candidate can have the same higher counting number. This is illustrated in Figure 10, where reference indicators “1” and “2” both appear four times. This could be solved, for example, by selecting the reference indicators according to a predefined position scheme. For example, when the reference indicators are represented by the numbers 0-2, as illustrated in Figures 9 and 10, the reference indicator represented by the greater, or lesser, number could be selected as predicted.
[0074] The difference between a block of predicted pixel values and the original pixel values, i.e. the font before being encoded, is encoded by transform encoding, eg DCT (Discrete Cosine Transform). The transform output comprises transform coefficients, which are then quantized. The number of transform coefficients associated with a block reflects how well the match between the predicted block and the general block is for the given quantization. Relatively few transform coefficients indicate that a good match exists. Consequently, benchmarks associated with blocks having few transform coefficients could be considered more reliable.
[0075] Thus, the count of occurrences could also be weighted according to, for example, the encoded transform coefficients associated with the reference blocks. As described above, a reference index associated with a neighboring block with few encoded transform coefficients could be considered a more reliable prediction than a reference index associated with a neighboring block with many transform coefficients, and a weight can thus be assigned. higher for benchmark forecasting. In another example, blocks with coded coefficients may have greater weight than blocks without coded coefficients, that is, skipped blocks. In another example, a reference index associated with a neighboring block that has a large MV partition, for example, a large MB, could be considered more reliable than a neighboring block with a smaller MV partition, and a larger weight would thus be designated. for benchmark forecasting. Using weights that are multiples of 2 is beneficial in view of the complexity. Weighted counting could also be implemented through the use of a lookup table.
[0076] Some benchmarks may be more related to each other than others. For example, when using coded reference information together, the reference indicator representing bi-forecast using the reference frames ref0 and ref1 will be more related to the reference indicator representing uni-forecast using one of ref0 and ref1, which, for example, to a benchmark representing uni-forecast using reference frame ref2. Thus, when counting a bi-forecast indicator, the corresponding indicators representing uni-forecast using the same frame of reference could be updated with some lower count value, ie, less than the count value for a “full match”. Similarly, benchmarks representing uni-forecast using, for example, ref0 and ref1, are more related to the corresponding benchmark representing bi-forecast using ref0 and ref1 than other bi-forecast benchmarks. Thus, when counting a one-way reference indicator, the reference indicator count corresponding to a multi-forecast where the reference frame in question is used can also be updated with some smaller value.
[0077] Figure 10 illustrates another exemplary mode of determining the frequency of occurrence of certain reference indicators for a current block, by counting the reference indicators associated with the neighboring blocks of the current block. Here, a current block is a large MB, and neighboring blocks are smaller in size than the current block. In some cases, it may be of interest to have the same number of neighboring blocks in the context regardless of the block size of the blocks in question.
[0078] An advantage of introducing benchmark prediction, or benchmark prediction, is to allow more efficient mapping of a VLC table. By considering the forecast and the VLC table together, more compression can be achieved. For example, when assuming 3-index encoding, eg (0,1,2), without using prediction, a fixed VLC table can be assigned, as illustrated in Figure 11. Assuming the index symbol “2 ” occurs most often, the table illustrated in Figure 11 would have a sub-ideal design, since “2” is encoded using a two-bit codeword, ie, “11”, while the less frequent “0” is encoded using one bit, ie, “0”.
[0079] When the prediction is added, a better design of the VLC table is enabled. An example of such an improved VLC table design is illustrated in Figure 12. In such an improved VLC design, the bits spent for encoding a reference indicator or index symbol can be adapted based on the prediction and so on the context of the current block. In the table illustrated in Figure 12, the reference indicator most frequently occurring in the context of the current block is encoded using a single-bit codeword, in this example, "0". The code words "10", and "11", comprising two bits, could be defined to identify, for example, the reference indicator having the second highest frequency of occurrence and the reference indicator having a third highest frequency of occurrence, respectively. Both the encoder and decoder of reference indicators must be aware of, and agree on, how to perform the prediction and how to interpret the code words.
[0080] The example described above is just a simple example, and it should be noted that the possible design is not limited to this one. There are several ways to assign different VLC tables to reference indicators or index symbols, for example when more reference indicators or index symbols are involved. An example approach could be to vary the indexing with the probability of occurrence of the indices, so that a frequently occurring reference indicator is assigned a low index number, and vice versa, and that a low index number costs fewer bits to encode than a high index number. Context Adaptive Binary Arithmetic Coding (CABAC) can be used to obtain variable bit cost to represent benchmarks or indices according to their probability. Some examples of different contexts are, for example, the fiducials associated with the neighboring blocks, a counting number of fiducials, or a weighted counting number of fiducials, as described above.
[0081] In the prior art, for example, using H.264, the generation of a bi-prediction block using two MVs/reference images, blocks or areas, involves an average over the two reference areas. When MV points to a sub-pel (sub-pixel) position in a reference area, the pixel values of the sub-pel position need to be generated first, before the average. The generation of pixel values from the sub-pel position is referred to as "spatial filtering", i.e. the generation process involves the spatial filtering of the respective reference areas. Thus, the prior art process for generating a bi-forecast block using two reference areas involves spatial filtering of the first area; spatial filtering of the second area, and finally averaged over the filtered areas. Spatial filtering is relatively demanding in terms of computational complexity.
[0082] It is perceived that this computational complexity could be reduced, which will be described below. In order to reduce complexity, a block can first be constructed based on the entire motion, for example by adding the two reference blocks together (without performing spatial filtering). This addition is an operation that is relatively inexpensive in terms of computational complexity. Then, the resulting block can be filtered, for example, interpolated, in order to obtain, for example, half or quarter-pel of the resolution. The sub-pel adjustment can be performed according to one of the MVs, or based on, for example, additional information encoded/decoded separately.
[0083] When a block is associated with more than one MV and the reference index, which is referred to herein as "multi-forecast", the respective one-way prediction components of the multi-forecast can be determined. Uni-forecast can be referred to as “single-forecast”, as can also, for example, intraforecast. It is realized that the partition of information could be derived based on the absolute difference between these one-way predictions. Partitioning information could be derived from both the encoder and the decoder in order to avoid overhead when transmitting fine-grained partition information.
[0084] In regions where the absolute difference between one-way predictions is relatively greater, a single one-way prediction or a bi-spatial prediction could be used. The only one-way prediction could be made according to the reference index and MV indicated in the bit stream for one of the bi-forecast (or multi-forecast) components of the bi-forecast. In other regions of the block, where the absolute difference between the one-way predictions is relatively smaller, bi-prediction can be used as indicated in the bit stream for the block. The decision to use a single bi-forecast / bi-forecast for a region, or to use the bi-forecast indicated in the bit stream, could be based, for example, on a comparison of the absolute difference between the unidirectional predictions associated with the region. and a predefined limit.
[0085] Assuming a bi-forecast block associated with 2 MVs and 2 reference areas. Conventionally, at this stage, this block is not split yet, but encoded as-is. However, it is clear that the “implicit” information obtained from the analysis of absolute differences or the “difference map” could be used to divide the block into more partitions, both in the encoder and in the decoder.
[0086] When the absolute difference of 2 reference areas or forecasts is calculated, there will be some region(s) in the difference map with the highest absolute value(s) and some region(s) (s) with the lowest absolute value(s). A low absolute difference value in one region generally means that the same object is represented in this region in both reference areas. If different objects would be represented in the region in the respective reference areas, the absolute difference would be large. If the same object is represented in a corresponding region in the respective reference areas, it is suitable and suitable for averaging the regions. If the corresponding regions represent different objects, it makes no sense to average them.
[0087] For example, a threshold could be defined, where difference values higher than the threshold represent "different regions of objects", and difference values less than the threshold represent "some regions of the object". The block could be partitioned according to these regions, according to a predefined scheme. As stated previously, partitioning could be performed based on implicit information, that is, without explicit signaling describing partitioning. Another advantage of this is that “non-square partitioning” can be supported. For example, when half of a ball is represented in a block, the partition of the block could be made very precise around the edge of the ball.
[0088] The encoder can signal to the decoder whether the partition approach described above is to be used. When signaled that the partition approach should be used, the encoder can optionally signal, for regions having a relatively high absolute difference value, which of the one-way predictions to use or which spatial bi-predictions to use. For example, weighted bi-forecast (different from the mean and possibly with DC offset) could be used. In some cases, it may be necessary to encode/decode some additional information to determine local parameters to be able to produce the spatial bi-prediction. The partition information obtained can also be used for prediction of partition information and the encoder can encode changes compared to the partition predicted to be decoded and used by the decoder. Deriving splitting information based on the difference between the reference areas can give an approximate indication of how the splitting should be done. Further refinement by sending refinements to the predicted partition information is also possible.
[0089] An example to obtain the partition information is to divide the block into 4 sub-blocks of equal size. The subblock with the largest normalized SAD ((Sum of absolute differences) (divided by the number of pixels it was calculated on)) normalized is iteratively divided into 4 regions of equal size if the normalized SAD of the subblock is, for example , equal to or greater than the normalized SAD of the “main” block 4 times larger. Normalized SAD refers to SAD per pixel or SAD by a specific sub-block size. Instead of SAD, other different pixel metrics could alternatively be used. An example is a metric with more weight in the strong local image structure, eg borders/lines. A remaining sub-block, which is not yet split, is then defined to be the partition that should use, for example, some bi-forecast modification.
[0090] Figure 13 shows an exemplary embodiment of the partition method. The block on the left side, 1302:a, is bi-predicted. SAD calculation is performed on the block (now denoted 1302:b), and the high SAD areas are identified and selected out, and are treated accordingly. In this example, the high SAD area is handled by switching to one-way prediction with only MV behind. Thus, the overall block can be partitioned into two partitions, one of which uses the bi-forecast indicated in the bitstream, and one (illustrated as comprising circles) uses uni-forecast (one of the components of the bi-forecast). Rate Distortion Optimization (RDO) could be used to select the best uni-forecast (component of the bi-forecast).
[0091] Another example of how to obtain partition information is to divide, for example, a bi-forecast block into a number of, for example, equal sizes of sub-blocks; determine the maximum SAD of the sub-size of the block in question, and select the sub-blocks having a “closed” SAD to, for example, within a certain range from this maximum value, to be part of a region that is to use some modified version of bi-prediction, or a one-way prediction.
[0092] In addition to partitioning, this approach can be used, for example, to determine the previously described RIS index or prior art reference indices, when a bi-prediction mode is used. For example, a smooth difference map for a region might suggest, and be interpreted as, that the region is possibly associated with a “bi-RIS index”. The approach could also be used as an alternative prediction or in combination with the Benchmark Prediction Indicator previously described. Selection can be made in both the encoder and the decoder based on SAD among possible bi-prediction candidates to select the combination with minimum SAD.
[0093] It should be noted that with the above described multi-forecasting based on approximate partitioning, rather than deriving block-based partitioning, other types of partitioning could be derived both at the encoder and at the decoder. This includes linear (e.g. horizontal, vertical, or diagonal) or non-linear partitioning of a block into two or more partitions, for example according to non-linear image processing methods such as edge detection and/or segmentation. For example, the multi-prediction difference signal can be segmented according to an image segmentation method such as edge detection or crescent region, and then block partition is derived based on the segmented difference signal.
[0094] The number of sub-partitions could be derived through any image processing methods such as image segmentation, or it could be signaled from encoder to decoder. As an alternative to linear or non-linear partitioning, pixel-based partitioning can also be applied. One variant would be to signal from the encoder to the decoder which partitioning method is used, another variant would be that the partitioning method is agreed between the encoder and the decoder through other signaling means. The advantage with multi-prediction based methods is that partition information can be derived based on information that is already available in the encoder and decoder, ie it does not have to be explicitly signaled, thus reducing the number of bits used for codification.
[0095] It should be noted that, according to partitioning based on multi-prediction, instead of switching from bi-prediction to uni-prediction with unidirectional VMs derived from the VMs used for bi-prediction, it is also possible to signal MVs and/or additional prediction modes (one-way inter-picture prediction, bi-directional inter-picture prediction, or intra-picture prediction) for sub-partitions. In other words, the number and shapes of partitions for a block could either be explicitly signaled and/or derived from implicit information, based on, for example, a segmentation method. Furthermore, VMs and/or predictive mode may be flagged by any or all of the resulting sub-partitions.
[0096] While the procedure as suggested above has been described with reference to the specific modalities provided by way of example, the description is generally only intended to illustrate the inventive concept and should not be taken as limiting the scope of the suggested methods and arrangements, which are defined by the appended claims . While described in general terms, the methods and provisions can be applied, for example, to different types of communication systems, using commonly available communication technologies, such as, for example, GSM/EDGE, WCDMA or LTE technologies or broadcast by satellite, terrestrial, or cable, eg DVB-S, DVB-T, or DVB-C, but also for storing/retrieving video to/from memory.
[0097] It should also be understood that the choice of interaction unit or modules, as well as the naming of units are for exemplifying purposes only, and suitable video manipulation entities to perform any of the methods described above can be configured in a plurality of alternative ways in order to be able to carry out the processes of the suggested actions.
[0098] It should be noted that the units or modules described in this report should be considered as logical entities and not necessarily as separate physical entities. Abbreviations
[0099] AVC Advanced Video Coding CABAC Adapted Binary Context Arithmetic Coding GOP Picture Group MB Block Macro MV Motion Vector RIS Reference Index Signaling/ SAD Indicator Reference Signaling Information Absolute Difference Sum VLC Variable Length Coding
权利要求:
Claims (28)
[0001]
1. Method in a video decoder for decoding information, the method characterized in that it comprises: obtaining (402) a single syntax element associated with a Be encoded block, wherein the single syntax element represents an entry in a first list of predefined reference and wherein the first list comprises one or more entries identifying at least one of a plurality of reference pictures and a single reference picture; identifying (404) a reference mode and one or more reference images based on the obtained syntax element; and decoding (406) the block Be based on the identified reference mode and one or more reference pictures, thereby providing a decoded block, B, of pixels.
[0002]
2. Method according to claim 1, characterized in that the identification of a reference mode and the one or more reference images are based on a predefined mapping between the obtained syntax element and the reference mode and a or more specific reference images to be used when decoding the Be block.
[0003]
3. Method according to any one of claims 1 or 2, characterized by the fact that each entry in the first list further identifies a reference mode.
[0004]
4. Method according to any one of claims 1 to 3, characterized in that the syntax element still represents a reference mode and an entry in a second predefined reference list.
[0005]
5. Method according to claim 1, characterized in that one or more reference figures include a past reference figure and a future reference figure.
[0006]
6. Arrangement (500) in a video decoder comprising a processor and a memory, the memory containing instructions executable by the processor, the arrangement characterized in that it is operated to: obtain a single syntax element associated with a Be encoded block, in that the single syntax element represents an entry in a first predefined reference list and wherein the first list comprises one or more entries identifying at least one of a plurality of reference pictures and a single reference picture; identify a reference mode and the one or more reference images based on the obtained syntax element; and decoding the B block based on the identified reference mode and one or more reference pictures, thus providing a B decoded block of pixels.
[0007]
7. Arrangement according to claim 6, characterized in that it is further operated to identify the reference mode and the one or more reference images based on a predefined mapping between the obtained syntax element and the reference mode and the one or more specific reference images to be used when decoding the Be block.
[0008]
8. Arrangement, according to any one of claims 6 or 7, characterized by the fact that each entry in the first list still identifies a reference mode.
[0009]
9. Arrangement according to claim 6, characterized in that it is further operated to interpret the syntax element as further representing a mode of reference and an entry in a second predefined reference list.
[0010]
10. Arrangement according to claim 9, characterized in that the second list comprises one or more entries identifying a unique reference image.
[0011]
11. Arrangement according to claim 6, characterized in that the one or more list entries are identified by a list index.
[0012]
12. Arrangement according to any one of claims 6, characterized in that it is further operated to interpret the number of bits representing the obtained syntax element as correlated to the probability of the specific values of the syntax element, in such a way that few bits imply probable values, and more bits imply less probable values.
[0013]
13. Arrangement according to claim 6, characterized in that it is further operated to perform prediction of reference information for Be, based on unique syntax elements associated with neighboring blocks.
[0014]
14. Arrangement according to claim 6, characterized in that it is further operated to identify one or more subregions of a block associated with multi-forecast, so that subregions of the respective corresponding regions of the multi reference blocks -forecast have a low correlation between them.
[0015]
15. Arrangement according to claim 6, characterized in that one or more reference figures include a past reference figure and a future reference figure.
[0016]
16. Computer-readable non-transient medium for controlling a video decoder, computer-readable non-transient medium characterized by comprising instructions that, when executed in the video decoder, cause the video decoder to: obtain a unique associated syntax element to a Be coded block, wherein the single syntax element represents an entry in a predefined first reference list and wherein the first list comprises one or more entries that identify at least one of a plurality of reference pictures and a single picture of reference; identify a reference mode and the one or more reference images based on the obtained syntax element; and decoding the block B based on the identified reference mode and the one or more reference pictures, providing a decoded block, B, of pixels.
[0017]
17. Method in a video encoder for encoding information, the method characterized in that it comprises the steps of: encoding (602) a block B of pixels using a reference mode and one or more reference pictures, thereby providing an encoded block Be; derive (604) a single syntax element identifying the reference mode and the one or more reference images, where the syntax element represents an entry in a first predefined reference list and where the first list comprises one or more entries identifying at least one of a plurality of reference images and a single reference image; - provide (606) the syntax element to a Be block decoder.
[0018]
18. Method according to claim 17, characterized in that the syntax element is derived by the reference mode used and the one or more reference images are mapped to the syntax element according to a predefined mapping scheme .
[0019]
19. Method according to claim 17, characterized in that each entry in the first list further identifies a mode of reference.
[0020]
20. Arrangement (700) in a video encoder comprising a processor and a memory, the memory containing instructions executable by the processor, the arrangement characterized in that it is operated to: encode a B block of pixels using a reference mode and one or more reference pictures, thus providing a Be coded block; derive a single syntax element identifying the mode of reference and the one or more reference images, where the syntax element represents an entry in a first predefined reference list, and where the first list comprises one or more entries identifying at least one of a plurality of reference images and a single reference image; and supply the syntax element to a Be block decoder.
[0021]
21. Arrangement according to claim 20, characterized in that it is further operated to derive the syntax element from a predetermined mapping between the reference mode and one or more reference images and the syntax element.
[0022]
22. Arrangement according to claim 20, characterized in that each entry in the first list further identifies a mode of reference.
[0023]
23. Arrangement according to claim 20, characterized in that it is further operated to derive the syntax element in such a way as to still represent a mode of reference and an entry in a second predefined reference list.
[0024]
24. Arrangement according to claim 23, characterized in that the second list comprises one or more entries identifying a respective unique reference image.
[0025]
25. Arrangement according to claim 20, characterized in that it is further operated to derive the syntax element by selecting a list index identifying one or more entries in one or more predefined reference lists.
[0026]
26. Arrangement according to claim 20, characterized in that it is further operated to select the number of bits representing the syntax element in such a way as to be correlated with the probability of the specific modes and images, which the syntax element identifies, so that higher probability corresponds to fewer bits, and lower probability corresponds to more bits.
[0027]
27. Arrangement according to claim 20, characterized in that it is further operated to perform prediction of reference information for B or Be, based on unique syntax elements associated with neighboring blocks.
[0028]
28. Computer-readable non-transient medium for controlling a video encoder, computer-readable non-transient medium characterized by comprising instructions which, when executed in the video encoder, make the video encoder: encode a block B of pixels using a mode and one or more reference pictures, thus providing a Be coded block; derive a single syntax element identifying the reference mode and the one or more reference images, where the syntax element represents an entry in a predefined first reference list, and where the first list comprises one or more entries that identify by at least one of a plurality of reference images and a single reference image; and provide the syntax element for a Be block decoder.
类似技术:
公开号 | 公开日 | 专利标题
BR112012015001B1|2021-06-29|METHOD IN A VIDEO DECODER TO DECODE INFORMATION, ARRANGEMENT IN A VIDEO DECODER, NON TRANSIENT COMPUTER READABLE MEDIUM TO CONTROL A VIDEO DECODER, METHOD IN A VIDEO ENCODER TO ENCODE INFORMATION, DEVICE IN A VIDEO ENCODER COMPUTER-READABLE NO TRANSIENT TO CONTROL A VIDEO ENCODER
US10609380B2|2020-03-31|Video encoding and decoding with improved error resilience
US11109050B2|2021-08-31|Video encoding and decoding
CN110677669B|2021-12-07|LUT with LIC
KR20100080792A|2010-07-12|Methods and apparatus for video encoding and decoding geometrically partitioned bi-predictive mode partitions
US10574863B2|2020-02-25|Video encoding and decoding
US20140321544A1|2014-10-30|Method and Device for Encoding a Sequence of Images and Method and Device for Decoding a Sequence of Images
US20190281287A1|2019-09-12|Method and Device for Encoding a Sequence of Images and Method and Device for Decoding a Sequence of Image
BR112019028012A2|2020-07-07|device and method for determining a motion vector, video encoder and decoder to encode and decode a plurality of photos
US11202076B2|2021-12-14|Method and apparatus for encoding or decoding video data in FRUC mode with reduced memory accesses
CN111064964B|2021-02-26|Encoding and decoding method, device and equipment
CN110944198A|2020-03-31|Chroma mode intra coding
同族专利:
公开号 | 公开日
US20120250766A1|2012-10-04|
RU2012130147A|2014-03-20|
US20120250765A1|2012-10-04|
EP2514209A4|2014-03-26|
WO2011075071A1|2011-06-23|
US9071845B2|2015-06-30|
RU2551207C2|2015-05-20|
US20150264393A1|2015-09-17|
CA2784515A1|2011-06-23|
CA2784517A1|2011-06-23|
WO2011075072A1|2011-06-23|
EP2514209A1|2012-10-24|
US10542284B2|2020-01-21|
JP5855013B2|2016-02-09|
CN102668561A|2012-09-12|
US9100658B2|2015-08-04|
BR112012015001A2|2016-11-29|
EP2514210A4|2014-03-19|
EP2514210A1|2012-10-24|
JP2013514718A|2013-04-25|
CN102656890B|2016-01-06|
JP2013514719A|2013-04-25|
CN102656890A|2012-09-05|
引用文献:
公开号 | 申请日 | 公开日 | 申请人 | 专利标题

US6535558B1|1997-01-24|2003-03-18|Sony Corporation|Picture signal encoding method and apparatus, picture signal decoding method and apparatus and recording medium|
WO2000064058A1|1999-04-16|2000-10-26|Fujitsu Limited|Encoder and decoder|
US7474699B2|2001-08-28|2009-01-06|Ntt Docomo, Inc.|Moving picture encoding/transmission system, moving picture encoding/transmission method, and encoding apparatus, decoding apparatus, encoding method decoding method and program usable for the same|
US7643559B2|2001-09-14|2010-01-05|Ntt Docomo, Inc.|Coding method, decoding method, coding apparatus, decoding apparatus, image processing system, coding program, and decoding program|
US6950469B2|2001-09-17|2005-09-27|Nokia Corporation|Method for sub-pixel value interpolation|
WO2003047270A1|2001-11-30|2003-06-05|Ntt Docomo, Inc.|Moving picture coding apparatus, moving picture decoding apparatus, moving picture coding method, moving picture decoding method, program, and computer-readable recording medium containing the program|
JP4015934B2|2002-04-18|2007-11-28|株式会社東芝|Video coding method and apparatus|
WO2004039083A1|2002-04-26|2004-05-06|Ntt Docomo, Inc.|Image encoding device, image decoding device, image encoding method, image decoding method, image encoding program, and image decoding program|
US7154952B2|2002-07-19|2006-12-26|Microsoft Corporation|Timestamp-independent motion vector prediction for predictive and bidirectionally predictive pictures|
AU2003242037A1|2002-07-02|2004-01-23|Matsushita Electric Industrial Co., Ltd.|Image encoding method and image decoding method|
KR100865034B1|2002-07-18|2008-10-23|엘지전자 주식회사|Method for predicting motion vector|
US7801217B2|2002-10-01|2010-09-21|Thomson Licensing|Implicit weighting of reference pictures in a video encoder|
JP4756573B2|2002-12-04|2011-08-24|トムソンライセンシング|Video cross fade encoder and coding method using weighted prediction|
WO2004064255A2|2003-01-07|2004-07-29|Thomson Licensing S.A.|Mixed inter/intra video coding of macroblock partitions|
KR100510136B1|2003-04-28|2005-08-26|삼성전자주식회사|Method for determining reference picture, moving compensation method thereof and apparatus thereof|
US7961786B2|2003-09-07|2011-06-14|Microsoft Corporation|Signaling field type information|
US7724827B2|2003-09-07|2010-05-25|Microsoft Corporation|Multi-layer run level encoding and decoding|
US7400681B2|2003-11-28|2008-07-15|Scientific-Atlanta, Inc.|Low-complexity motion vector prediction for video codec with two lists of reference pictures|
US8036271B2|2004-02-24|2011-10-11|Lsi Corporation|Method and apparatus for determining a second picture for temporal direct-mode block prediction|
JP4575129B2|2004-12-02|2010-11-04|ソニー株式会社|DATA PROCESSING DEVICE, DATA PROCESSING METHOD, PROGRAM, AND PROGRAM RECORDING MEDIUM|
EP1913778A4|2005-07-08|2010-04-28|Lg Electronics Inc|Method for modeling coding information of a video signal to compress/decompress the information|
CN101375594B|2006-01-12|2011-09-07|Lg电子株式会社|Processing multiview video|
CN101379835B|2006-02-02|2011-08-24|汤姆逊许可公司|Method and apparatus for motion estimation using combined reference bi-prediction|
US8270492B2|2006-05-12|2012-09-18|Panasonic Corporation|Moving picture decoding device|
KR100823287B1|2007-01-03|2008-04-21|삼성전자주식회사|Method and apparatus for encoding and decoding multi-view image based on global disparity vector|
WO2008154041A1|2007-06-14|2008-12-18|Thomson Licensing|Modifying a coded bitstream|
US20090187960A1|2008-01-17|2009-07-23|Joon Hui Lee|IPTV receiving system and data processing method|
US20100086051A1|2008-10-06|2010-04-08|Lg Electronics Inc.|Method and an apparatus for processing a video signal|US20110310976A1|2010-06-17|2011-12-22|Qualcomm Incorporated|Joint Coding of Partition Information in Video Coding|
JP2012023651A|2010-07-16|2012-02-02|Sony Corp|Image processing device and image processing method|
CA2806523C|2010-07-28|2016-07-05|Nokia Corporation|Method and apparatus for providing complexity balanced entropy coding|
US9398308B2|2010-07-28|2016-07-19|Qualcomm Incorporated|Coding motion prediction direction in video coding|
US9025661B2|2010-10-01|2015-05-05|Qualcomm Incorporated|Indicating intra-prediction mode selection for video coding|
WO2012070239A1|2010-11-26|2012-05-31|パナソニック株式会社|Image encoding method, image decoding method, image encoding device, image decoding device, program, and integrated circuit|
US8913662B2|2011-01-06|2014-12-16|Qualcomm Incorporated|Indicating intra-prediction mode selection for video coding using CABAC|
US9008176B2|2011-01-22|2015-04-14|Qualcomm Incorporated|Combined reference picture list construction for video coding|
US9008181B2|2011-01-24|2015-04-14|Qualcomm Incorporated|Single reference picture list utilization for interprediction video coding|
CN103563382A|2011-03-11|2014-02-05|三星电子株式会社|Method and apparatus for encoding images and method and apparatus for decoding images|
US8934552B2|2011-03-31|2015-01-13|Qualcomm Incorporated|Combined reference picture list construction and mapping|
KR20120140592A|2011-06-21|2012-12-31|한국전자통신연구원|Method and apparatus for reducing computational complexity of motion compensation and increasing coding efficiency|
GB2492337B|2011-06-27|2018-05-09|British Broadcasting Corp|Video encoding and decoding using reference pictures|
US9699456B2|2011-07-20|2017-07-04|Qualcomm Incorporated|Buffering prediction data in video coding|
JP2013126093A|2011-12-14|2013-06-24|Sony Corp|Image processing apparatus and image processing method|
WO2013108690A1|2012-01-19|2013-07-25|ソニー株式会社|Image processing device and method|
ES2780849T3|2012-01-19|2020-08-27|Vid Scale Inc|Method and apparatus for signaling and building video encoding reference picture lists|
US9762903B2|2012-06-01|2017-09-12|Qualcomm Incorporated|External pictures in video coding|
US9667984B2|2012-09-28|2017-05-30|Dolby International Ab|Image decoding device and image coding device|
CN103237212B|2013-04-08|2016-04-13|浙江大学|A kind of method and device generating the queue of enhancement layer block cell encoding information reference|
US9948934B2|2014-07-02|2018-04-17|Apple Inc.|Estimating rate costs in video encoding operations using entropy encoding statistics|
CN107079161B|2014-09-01|2020-11-20|寰发股份有限公司|Method for intra picture block copying for screen content and video coding|
US10148953B2|2014-11-10|2018-12-04|Samsung Electronics Co., Ltd.|System and method for intra prediction in video coding|
CA3025490A1|2016-05-28|2017-12-07|Mediatek Inc.|Method and apparatus of current picture referencing for video coding using affine motion compensation|
WO2018181252A1|2017-03-28|2018-10-04|Panasonic Intellectual Property Corporation Of America|Encoding apparatus, decoding apparatus, encoding method, and decoding method|
CN107094251B|2017-03-31|2021-07-23|浙江大学|Video and image coding and decoding method and device based on spatial position adaptive quality adjustment|
JP6825506B2|2017-07-19|2021-02-03|富士通株式会社|Moving image coding device, moving image coding method, moving image coding computer program, moving image decoding device and moving image decoding method, and moving image decoding computer program|
JP2018007279A|2017-10-11|2018-01-11|株式会社東芝|Decoding method and decoder|
JP2018007278A|2017-10-11|2018-01-11|株式会社東芝|Decoding method and decoder|
US20190320203A1|2018-04-13|2019-10-17|Mediatek Inc.|Implicit Transform Settings|
US11196988B2|2018-12-17|2021-12-07|Apple Inc.|Reference picture management and list construction|
法律状态:
2018-03-27| B15K| Others concerning applications: alteration of classification|Ipc: H04N 19/573 (2014.01), H04N 19/159 (2014.01), H04N |
2019-01-08| B06F| Objections, documents and/or translations needed after an examination request according [chapter 6.6 patent gazette]|
2020-01-07| B06U| Preliminary requirement: requests with searches performed by other patent offices: procedure suspended [chapter 6.21 patent gazette]|
2021-04-13| B09A| Decision: intention to grant [chapter 9.1 patent gazette]|
2021-06-29| B16A| Patent or certificate of addition of invention granted|Free format text: PRAZO DE VALIDADE: 20 (VINTE) ANOS CONTADOS A PARTIR DE 17/12/2010, OBSERVADAS AS CONDICOES LEGAIS. PATENTE CONCEDIDA CONFORME ADI 5.529/DF, , QUE DETERMINA A ALTERACAO DO PRAZO DE CONCESSAO. |
优先权:
申请号 | 申请日 | 专利标题
US28724209P| true| 2009-12-17|2009-12-17|
US61/287,242|2009-12-17|
PCT/SE2010/051412|WO2011075071A1|2009-12-17|2010-12-17|Method and arrangement for video coding|
[返回顶部]